1,711 research outputs found
Confident Kernel Sparse Coding and Dictionary Learning
In recent years, kernel-based sparse coding (K-SRC) has received particular
attention due to its efficient representation of nonlinear data structures in
the feature space. Nevertheless, the existing K-SRC methods suffer from the
lack of consistency between their training and test optimization frameworks. In
this work, we propose a novel confident K-SRC and dictionary learning algorithm
(CKSC) which focuses on the discriminative reconstruction of the data based on
its representation in the kernel space. CKSC focuses on reconstructing each
data sample via weighted contributions which are confident in its corresponding
class of data. We employ novel discriminative terms to apply this scheme to
both training and test frameworks in our algorithm. This specific design
increases the consistency of these optimization frameworks and improves the
discriminative performance in the recall phase. In addition, CKSC directly
employs the supervised information in its dictionary learning framework to
enhance the discriminative structure of the dictionary. For empirical
evaluations, we implement our CKSC algorithm on multivariate time-series
benchmarks such as DynTex++ and UTKinect. Our claims regarding the superior
performance of the proposed algorithm are justified throughout comparing its
classification results to the state-of-the-art K-SRC algorithms.Comment: 10 pages, ICDM 2018 conferenc
Tree Edit Distance Learning via Adaptive Symbol Embeddings
Metric learning has the aim to improve classification accuracy by learning a
distance measure which brings data points from the same class closer together
and pushes data points from different classes further apart. Recent research
has demonstrated that metric learning approaches can also be applied to trees,
such as molecular structures, abstract syntax trees of computer programs, or
syntax trees of natural language, by learning the cost function of an edit
distance, i.e. the costs of replacing, deleting, or inserting nodes in a tree.
However, learning such costs directly may yield an edit distance which violates
metric axioms, is challenging to interpret, and may not generalize well. In
this contribution, we propose a novel metric learning approach for trees which
we call embedding edit distance learning (BEDL) and which learns an edit
distance indirectly by embedding the tree nodes as vectors, such that the
Euclidean distance between those vectors supports class discrimination. We
learn such embeddings by reducing the distance to prototypical trees from the
same class and increasing the distance to prototypical trees from different
classes. In our experiments, we show that BEDL improves upon the
state-of-the-art in metric learning for trees on six benchmark data sets,
ranging from computer science over biomedical data to a natural-language
processing data set containing over 300,000 nodes.Comment: Paper at the International Conference of Machine Learning (2018),
2018-07-10 to 2018-07-15 in Stockholm, Swede
Dimensionality Reduction Mappings
A wealth of powerful dimensionality reduction methods has been established which can be used for data visualization and preprocessing. These are accompanied by formal evaluation schemes, which allow a quantitative evaluation along general principles and which even lead to further visualization schemes based on these objectives. Most methods, however, provide a mapping of a priorly given finite set of points only, requiring additional steps for out-of-sample extensions. We propose a general view on dimensionality reduction based on the concept of cost functions, and, based on this general principle, extend dimensionality reduction to explicit mappings of the data manifold. This offers simple out-of-sample extensions. Further, it opens a way towards a theory of data visualization taking the perspective of its generalization ability to new data points. We demonstrate the approach based on a simple global linear mapping as well as prototype-based local linear mappings.
- …